NVIDIA Enhances AI Scalability with NIM Operator 3.0.0 Release
NVIDIA has launched NIM Operator 3.0.0, a significant upgrade aimed at optimizing AI inference scalability within Kubernetes environments. The release introduces multi-LLM and multi-node capabilities, enabling efficient deployment of large language models across distributed GPU clusters.
Collaboration with Red Hat enhances integration with KServe, streamlining lifecycle management for AI microservices. The update supports diverse AI applications—from reasoning to speech processing—while improving GPU utilization for cost-effective scaling.